Compare Page

Data freshness

Characteristic Name: Data freshness
Dimension: Currency
Description: Data which is subjected to changes over the time should be fresh and up-to-date with respect to its intended use.
Granularity: Element
Implementation Type: Process-based approach
Characteristic Type: Usage

Verification Metric:

The number of tasks failed or under performed due to lack of data freshness
The number of complaints received due to lack of data freshness

GuidelinesExamplesDefinitons

The implementation guidelines are guidelines to follow in regard to the characteristic. The scenarios are examples of the implementation

Guidelines: Scenario:
Identify the natural factors which creates a particular data obsolete (1) A seasonal change may impact the customer's food preferences. (2) Customers who are students may change their addresses frequently.
Considering the above factors plan for data refreshing activities by specify the frequency of refreshing the data elements and adhere to the plan. (1) Customer contact information should be refreshed annually.
Identify the master data that may change over the time but may be used in longitudinal analysis. (1) Name of customer in 2001 is ABC (PLC) Ltd, after a merger in 2006 its name is XYZ (PLC). This customer is an ongoing customer in the customer master file
For such master data maintain longitudinal versions with time a stamp in such a way they can be linked in longitudinal analysis (1) 2001-2005: ABC (PLC) (2) 2006-20012: XYZ (PLC)

Validation Metric:

How mature is the process for ensuring data freshness

These are examples of how the characteristic might occur in a database.

Example: Source:
let us consider two databases, say A and B, that contain the same data. If at time t a user updates data in database A and another user reads the same data from database B at time t' (t < t' ), the latter will read incorrect data. If t and f are included within the time interval between two subsequent data realignments C. Cappiello, C. Francalanci, and B. Pernici, “Time-Related Factors of Data Quality in Multichannel Information System” in Journal of Management Information Systems, Vol. 20, No. 3, M.E. Sharpe, Inc., 2004, pp.71-91.
currency indicates how stale is the account balance presented to the user with respect to the real balance at the bank database. V. Peralta, “Data Freshness and Data Accuracy: A State of The Art”, Instituto de Computacion, Facultad de Ingenieria, Universidad de la Republica, Uruguay, Tech. Rep. TR0613, 2006.
Consider an air traffic control center which receives data from several controller stations. To regulate air traffic, the traffic control center has to cope with uncertain data.Thus, the decision process must balance the delaying receiving more accurate data of airplane positions and the critical period of time in which an“effective” decision must be made to regulate traffic; B. Pernici, “Advanced Information Systems Engineering” in proc. The 22nd International Conference, CAiSE, Hammamet, Tunisia, June 2010.

The Definitions are examples of the characteristic that appear in the sources provided.

Definition: Source:
A measure of the rate of negative change to the data. D. McGilvray, “Executing Data Quality Projects: Ten Steps to Quality Data and Trusted Information”, Morgan Kaufmann Publishers, 2008.
Is the information upto-date and not obsolete? EPPLER, M. J. 2006. Managing information quality: increasing the value of information in knowledge-intensive products and processes, Springer.
Data is accurate if it is up to date – anti example: “Current president of the USA: Bill Clinton”. KIMBALL, R. & CASERTA, J. 2004. The data warehouse ETL toolkit: practical techniques for extracting. Cleaning, Conforming, and Delivering, Digitized Format, originally published.
Currency refers to the degree to which information is current with the world that it models. Currency can measure how up to date information is and whether is it correct despite possible time-related changes. Timeliness refers to the time. LOSHIN, D. 2001. Enterprise knowledge management: The data quality approach, Morgan Kaufmann Pub.
Currency refers to the degree to which information is current with the world that it models. Currency can measure how “up-to-date” information is, and whether it is correct despite possible time-related changes. Data currency may be measured as a function of the expected frequency rate at which different data elements are expected to be refreshed, as well as verifying that the data is up to date. For example, one might assert that the contact information for each customer must be current, indicating a requirement to maintain the most recent values associated with the individual’s contact data. LOSHIN, D. 2006. Monitoring Data quality Performance using Data Quality Metrics. Informatica Corporation.
A datum value is up-to-date if it is correct in spite of a possible discrepancy caused by time related change to the correct values; a datum is outdate at time t if it is incorrect at t but was correct at some time preceding t. currency refers to a degree to which a datum in question is up-to-date. REDMAN, T. C. 1997. Data quality for the information age, Artech House, Inc.

 

Statistical validity

Characteristic Name: Statistical validity
Dimension: Validity
Description: Computed data must be statistically valid
Granularity: Information object
Implementation Type: Process-based approach
Characteristic Type: Usage

Verification Metric:

The number of tasks failed or under performed due to lack of statistical validity in data
The number of complaints received due to lack of statistical validity of data

GuidelinesExamplesDefinitons

The implementation guidelines are guidelines to follow in regard to the characteristic. The scenarios are examples of the implementation

Guidelines: Scenario:
Establish the population of interest unambiguously with appropriate justification (maintain documentation) (1) Both credit customers and cash customers are considered for a survey on customer satisfaction.
Establish an appropriate sampling method with appropriate justification (1) Stratified sampling is used to investigate drug preference of the medical officers
Establish statistical validity of samples -avoid over coverage and under coverage (maintain documentation) (1) Samples are taken from all income levels in a survey on vaccination
Maintain consistency of samples in case longitudinal analysis is performed. (Maintain documentation) (1) Same population is used over the time to collect epidemic data for a longitudinal analysis
Ensure that valid statistical methods are used to enable valid inferences about data, valid comparisons of parameters and generalise the findings. (1) Poisson distribution is used to make inferences since data generating events are occurred in a fixed interval of time and/or space
Ensure that the acceptable variations for estimated parameters are established with appropriate justifications (1) 95% confidence interval is used in estimating the mean value
Ensure that appropriate imputation measures are taken to nullify the impact of problems relating to outliers, data collection and data collection procedures and the edit rules are defined and maintained. (1) Incomplete responses are removed from the final data sample

Validation Metric:

How mature is the process to maintain statistical validity of data

These are examples of how the characteristic might occur in a database.

Example: Source:
if a column should contain at least one occurrence of all 50 states, but the column contains only 43 states, then the population is incomplete. Y. Lee, et al., “Journey to Data Quality”, Massachusetts Institute of Technology, 2006.

The Definitions are examples of the characteristic that appear in the sources provided.

Definition: Source:
Coherence of data refers to the internal consistency of the data. Coherence can be evaluated by determining if there is coherence between different data items for the same point in time, coherence between the same data items for different points in time or coherence between organisations or internationally. Coherence is promoted through the use of standard data concepts, classifications and target populations. HIQA 2011. International Review of Data Quality Health Information and Quality Authority (HIQA), Ireland. http://www.hiqa.ie/press-release/2011-04-28-international-review-data-quality.
1) Accuracy in the general statistical sense denotes the closeness of computations or estimates to the exact or true values.

2) Coherence of statistics is their adequacy to be reliably combined in different ways and for various uses.

LYON, M. 2008. Assessing Data Quality ,
Monetary and Financial Statistics.
Bank of England. http://www.bankofengland.co.uk/
statistics/Documents/ms/articles/art1mar08.pdf.